Local Convergence of Sequential Convex Programming for Nonconvex Optimization
نویسندگان
چکیده
where c ∈ R, g : R → R is non-linear and smooth on its domain, and Ω is a nonempty closed convex subset in R. This paper introduces sequential convex programming (SCP), a local optimization method for solving the nonconvex problem (P). We prove that under acceptable assumptions the SCP method locally converges to a KKT point of (P) and the rate of convergence is linear. Problems in the form of (P) conveniently formulate many problems of interest such as least squares problems, quadratically constrained quadratic programming, nonlinear semidefinite programming (SDP), and nonlinear second order cone programming problems (see, e.g., [1, 2, 5, 6, 10]). In nonlinear optimal control, by using direct transcription methods, the resulting problem is usually formulated as an optimization problem of the form (P) where the equality constraint g(x) = 0 originates from the dynamic system of an optimal control problem. The main difficulty of the problem (P) is concentrated in the nonlinear constraint g(x) = 0 that can be overcome by linearizing it around the current
منابع مشابه
On Sequential Optimality Conditions without Constraint Qualifications for Nonlinear Programming with Nonsmooth Convex Objective Functions
Sequential optimality conditions provide adequate theoretical tools to justify stopping criteria for nonlinear programming solvers. Here, nonsmooth approximate gradient projection and complementary approximate Karush-Kuhn-Tucker conditions are presented. These sequential optimality conditions are satisfied by local minimizers of optimization problems independently of the fulfillment of constrai...
متن کاملAccelerated gradient methods for nonconvex nonlinear and stochastic programming
In this paper, we generalize the well-known Nesterov’s accelerated gradient (AG) method, originally designed for convex smooth optimization, to solve nonconvex and possibly stochastic optimization problems. We demonstrate that by properly specifying the stepsize policy, the AG method exhibits the best known rate of convergence for solving general nonconvex smooth optimization problems by using ...
متن کاملA sequential parametric convex approximation method with applications to nonconvex truss topology design problems
We describe a general scheme for solving nonconvex optimization problems, where in each iteration the nonconvex feasible set is approximated by an inner convex approximation. The latter is defined using an upper bound on the nonconvex constraint functions. Under appropriate conditions on this upper bounding convex function, a monotone convergence to a KKT point is established. The scheme is app...
متن کاملA feasible second order bundle algorithm for nonsmooth, nonconvex optimization problems with inequality constraints: I. Derivation and convergence
This paper extends the SQP-approach of the well-known bundle-Newton method for nonsmooth unconstrained minimization to the nonlinearly constrained case. Instead of using a penalty function or a filter or an improvement function to deal with the presence of constraints, the search direction is determined by solving a convex quadratically constrained quadratic program to obtain good iteration poi...
متن کاملReal-Time Sequential Convex Programming for Optimal Control Applications
This paper proposes real-time sequential convex programming (RTSCP), a method for solving a sequence of nonlinear optimization problems depending on an online parameter. We provide a contraction estimate for the proposed method and, as a byproduct, a new proof of the local convergence of sequential convex programming. The approach is illustrated by an example where RTSCP is applied to nonlinear...
متن کامل